161 research outputs found

    Automatic Concept Extraction in Semantic Summarization Process

    Get PDF
    The Semantic Web offers a generic infrastructure for interchange, integration and creative reuse of structured data, which can help to cross some of the boundaries that Web 2.0 is facing. Currently, Web 2.0 offers poor query possibilities apart from searching by keywords or tags. There has been a great deal of interest in the development of semantic-based systems to facilitate knowledge representation and extraction and content integration [1], [2]. Semantic-based approach to retrieving relevant material can be useful to address issues like trying to determine the type or the quality of the information suggested from a personalized environment. In this context, standard keyword search has a very limited effectiveness. For example, it cannot filter for the type of information, the level of information or the quality of information. Potentially, one of the biggest application areas of content-based exploration might be personalized searching framework (e.g., [3],[4]). Whereas search engines provide nowadays largely anonymous information, new framework might highlight or recommend web pages related to key concepts. We can consider semantic information representation as an important step towards a wide efficient manipulation and retrieval of information [5], [6], [7]. In the digital library community a flat list of attribute/value pairs is often assumed to be available. In the Semantic Web community, annotations are often assumed to be an instance of an ontology. Through the ontologies the system will express key entities and relationships describing resources in a formal machine-processable representation. An ontology-based knowledge representation could be used for content analysis and object recognition, for reasoning processes and for enabling user-friendly and intelligent multimedia content search and retrieval. Text summarization has been an interesting and active research area since the 60’s. The definition and assumption are that a small portion or several keywords of the original long document can represent the whole informatively and/or indicatively. Reading or processing this shorter version of the document would save time and other resources [8]. This property is especially true and urgently needed at present due to the vast availability of information. Concept-based approach to represent dynamic and unstructured information can be useful to address issues like trying to determine the key concepts and to summarize the information exchanged within a personalized environment. In this context, a concept is represented with a Wikipedia article. With millions of articles and thousands of contributors, this online repository of knowledge is the largest and fastest growing encyclopedia in existence. The problem described above can then be divided into three steps: • Mapping of a series of terms with the most appropriate Wikipedia article (disambiguation). • Assigning a score for each item identified on the basis of its importance in the given context. • Extraction of n items with the highest score. Text summarization can be applied to many fields: from information retrieval to text mining processes and text display. Also in personalized searching framework text summarization could be very useful. The chapter is organized as follows: the next Section introduces personalized searching framework as one of the possible application areas of automatic concept extraction systems. Section three describes the summarization process, providing details on system architecture, used methodology and tools. Section four provides an overview about document summarization approaches that have been recently developed. Section five summarizes a number of real-world applications which might benefit from WSD. Section six introduces Wikipedia and WordNet as used in our project. Section seven describes the logical structure of the project, describing software components and databases. Finally, Section eight provides some consideration..

    Enabling smart learning systems within smart cities using open data

    Get PDF
    Deploying ad-hoc learning environments to use and represent data from multiple sources and networks and to dynamically respond to user demands could be very expensive and ineffective in the long run. Moreover, most of the available data is wasted without extracting potentially useful information and knowledge because of the lack of established mechanisms and standards. It is preferable to focus on data availability to choose and develop interoperability strategies suitable for smart learning systems based on open standards and allowing seamless integration of third-party data and custom applications. This paper highlights the opportunity to take advantage of emerging technologies, like the linked open data platforms and automatic reasoning to effectively handle the vast amount of information and to use data linked queries in the domain of cognitive smart learning systems

    The Digital Girls Response to Pandemic: Impacts of in Presence and Online Extracurricular Activities on Girls Future Academic Choices

    Get PDF
    In the last few years, several initiatives based on extracurricular activities have been organized in many countries around the world, with the aim to reduce the digital gender gap in STEM (Science, Technology, Engineering, Math) fields. Among them, the Digital Girls summer camp, organized every year since 2014 by two Italian universities with the aim to attract female students to ICT (Information and Communication Technologies) disciplines, represents quite a unique initiative for its characteristics of long-duration (3–4 entire weeks) and complete gratuitousness for the participants. The COVID-19 emergency imposed severe changes to such activities, that had to be modified and carried out in the online mode as a consequence of social distancing. However, on one hand, the general lack of high-quality evaluations of these initiatives hinders the possibility to understand the actual impact of extracurricular activities on the future academic choices of the participants. On the other hand, the availability of data collected over different editions of Digital Girls has allowed us to analyze the summer camp impact and to evaluate the pros and cons of in-presence and online activities. The main contribution of this paper is twofold. First, we present an overview of existing experiences, at the national (Italian) and international levels, to increase female participation in integrated STEM and ICT fields. Second, we analyze how summer camp participation can influence girls’ future academic choices, with specific attention to ICT-related disciplines. In particular, the collection of a significant amount of data through anonymous surveys conducted before and after the camp activities over the two editions allowed us to evidence the different impacts of in-presence and online extracurricular activitie

    A General Semantic Web Approach for Data Analysis on Graduates Statistics

    Get PDF
    Currently, several datasets released in a Linked Open Data format are available at a national and international level, but the lack of shared strategies concerning the definition of concepts related to the statistical publishing community makes difficult a comparison among given facts starting from different data sources. In order to guarantee a shared representation framework for what concerns the dissemination of statistical concepts about graduates, we developed SW4AL, an ontology- based system for graduate’s surveys domain. The developed system transforms low-level data into an enriched information model and is based on the AlmaLaurea surveys covering more than 90% of Italian graduates. SW4AL: i) semantically describes the different peculiarities of the graduates; ii) promotes the structured definition of the AlmaLaurea data and the following publication in the Linked Open Data context; iii) provides their reuse in the open data scope; iv) enables logical reasoning about knowledge representation. SW4AL establishes a common semantic for addressing the concept of graduate’s surveys domain by proposing the creation of a SPARQL endpoint and a Web based interface for the query and the visualization of the structured data

    NLG-Metricverse: An End-to-End Library for Evaluating Natural Language Generation

    Get PDF
    Driven by deep learning breakthroughs, natural language generation (NLG) models have been at the center of steady progress in the last few years, with a ubiquitous task influence. However, since our ability to generate human-indistinguishable artificial text lags behind our capacity to assess it, it is paramount to develop and apply even better automatic evaluation metrics. To facilitate researchers to judge the effectiveness of their models broadly, we introduce NLG-Metricverse—an end-to-end open-source library for NLG evaluation based on Python. Our framework provides a living collection of NLG metrics in a unified and easy-to-use environment, supplying tools to efficiently apply, analyze, compare, and visualize them. This includes (i) the extensive support to heterogeneous automatic metrics with n-arity management, (ii) the meta-evaluation upon individual performance, metric-metric and metric-human correlations, (iii) graphical interpretations for helping humans better gain score intuitions, (iv) formal categorization and convenient documentation to accelerate metrics understanding. NLG-Metricverse aims to increase the comparability and replicability of NLG research, hopefully stimulating new contributions in the area

    AN ADAPTIVE ASSESSMENT SYSTEM TO EVALUATE STUDENT ABILITY LEVEL

    Get PDF
    Abstract: The experience from years of development and use, the advance of technology, and the development of authoring tools for questions and tests has resulted in a sophisticated, computer based assessment system. However, there is still a lot of room for further development. Some of the current ideas for development are discussed in the remainder of this work. A primary aim of assessment, both formative and summative, is provide the necessary information to improve future educational experiences because it provides feedback on whether the course and learning objectives have been achieved to satisfactory level. Yet, it is important that the assessment data be accurate and relevant to effectively make informed decisions about the curriculum. Moreover, formative assessment can also be used to help bridge the gap between assessment and learning. This may be achieved particularly where assessment strategies are combined with useful feedback, and integrated within the learning process. The answers to the described objectives are enhanced if we could integrate adaptive testing techniques; accurate and fitted assessment data may improve both the curriculum and the student ability level. The idea behind a computerized adaptive testing (CAT) is quite forward: to apply to each examinee only those items useful to know his proficiency level. As a consequence of this, CAT is more efficient than conventional (i.e., fixeditem) tests. It provides more precise measurements for same-length tests or shorter tests for same-precision measurements

    A new digital divide threatening resilience: exploring the need for educational, firm-based, and societal investments in ICT human capital

    Get PDF
    The knowledge, skills, and abilities that human capital offers create tangible and intangible assets that equip organizations to thrive. In particular, in today’s Industry 4.0 environment, training, recruiting, and retaining highly qualified ICT-ready professionals remains a problem for many organizations including educational, governmental, healthcare, and business organizations. The COVID-19 pandemic revealed the importance of digital assets to our economies, and it is also demonstrating that there is potentially a new digital divide with even worse implications for companies, economies, and society, which is threatening the resilience of business, governance, and society. In this paper, we respond to the question “how can we develop ICT human capital in our global economy in an equitable, inclusive, and purposeful manner such that not organizations thrive, but also to promote social justice and equity in our global economy?

    CometAnalyser : A user-friendly, open-source deep-learning microscopy tool for quantitative comet assay analysis

    Get PDF
    Comet assay provides an easy solution to estimate DNA damage in single cells through microscopy assessment. It is widely used in the analysis of genotoxic damages induced by radiotherapy or chemotherapeutic agents. DNA damage is quantified at the single-cell level by computing the displacement between the genetic material within the nucleus, typically called ``comet head", and the genetic material in the surrounding part of the cell, considered as the ``comet tail". Today, the number of works based on Comet Assay analyses is really impressive. In this work, besides revising the solutions available to obtain reproducible and reliable quantitative data, we developed an easy-to-use tool named CometAnalyser. It is designed for the analysis of both fluorescent and silver-stained wide-field microscopy images and allows to automatically segment and classify the comets, besides extracting Tail Moment and several other intensity/morphological features for performing statistical analysis. CometAnalyser is an open-source deep-learning tool. It works with Windows, Macintosh, and UNIX-based systems. Source code, standalone versions, user manual, sample images, video tutorial and further documentation are freely available at: https://sourceforge.net/p/cometanalyser. (c) 2022 The Author(s). Published by Elsevier B.V. on behalf of Research Network of Computational and Structural Biotechnology. This is an open access article under the CC BY license (http://creativecommons. org/licenses/by/4.0/).Peer reviewe
    • …
    corecore